#### Token {: #token data-category=gen-ai }
A token is the smallest unit of text an LLM can work with. When processing a user prompt, the LLM splits the input text into tokens and generates the output token by token. Different LLMs can separate the same text into tokens differently as they learn their vocabulary from the data they were trained on. Depending on the language and the LLM, a token can be a character, a group of characters, a word, or any other unit of text. An approximation of token length for GPT models is `1 token ~= 4` chars in English.

